Neural Learning in Structured Parameter Spaces - Natural Riemannian Gradient
نویسنده
چکیده
The parameter space of neural networks has a Riemannian metric structure. The natural Riemannian gradient should be used instead of the conventional gradient, since the former denotes the true steepest descent direction of a loss function in the Riemannian space. The behavior of the stochastic gradient learning algorithm is much more effective if the natural gradient is used. The present paper studies the information-geometrical structure of perceptrons and other networks, and prove that the on-line learning method based on the natural gradient is asymptotically as efficient as the optimal batch algorithm. Adaptive modification of the learning constant is proposed and analyzed in terms of the Riemannian measure and is shown to be efficient. The natural gradient is finally applied to blind separation of mixtured independent signal sources.
منابع مشابه
Gradient Learning in Structured Parameter Spaces: Adaptive Blind Separation of Signal Sources
The present paper discusses the natural gradient descent learning rules in parameter spaces which have Riemannian geometrical structures. A modi cation is necessary for de ning the steepest descent (gradient) direction in a Riemannian parameter space. Parameter spaces of multilayer perceptrons are good examples of the Riemannian nature. Another example is the space of matrices on which adaptive...
متن کاملA Neural Stiefel Learning based on Geodesics Revisited
In this paper we present an unsupervised learning algorithm of neural networks with p inputs and m outputs whose weight vectors have orthonormal constraints. In this setting the learning algorithm can be regarded as optimization posed on the Stiefel manifold, and we generalize the natural gradient method to this case based on geodesics. By exploiting its geometric property as a quotient space: ...
متن کاملAdaptive Natural Gradient Learning Based on Riemannian Metric of Score Matching
The natural gradient is a powerful method to improve the transient dynamics of learning by considering the geometric structure of the parameter space. Many natural gradient methods have been developed with regards to Kullback-Leibler (KL) divergence and its Fisher metric, but the framework of natural gradient can be essentially extended to other divergences. In this study, we focus on score mat...
متن کاملKey Diagonal Blocks of the Fisher Information Matrix on Neural Manifold of Full-Parametrised Multilayer Perceptrons
Abstract: It’s well known the natural gradient learning (NGL) ([1]) may avoid global optima or phenomena of plateau in the training process since it takes into consideration the intrinsic geometric structure of the parameter space. But, natural gradient ([1]) is itself induced by Fisher information matrix (FIM) ([2]) defined on the 1-form tangent space ([3]), therefore calculation of relevant F...
متن کاملPractical Riemannian Neural Networks
We provide the first experimental results on non-synthetic datasets for the quasidiagonal Riemannian gradient descents for neural networks introduced in [Oll15]. These include the MNIST, SVHN, and FACE datasets as well as a previously unpublished electroencephalogram dataset. The quasi-diagonal Riemannian algorithms consistently beat simple stochastic gradient gradient descents by a varying mar...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1996